1,604 research outputs found
New Physics Agnostic Selections For New Physics Searches
We discuss a model-independent strategy for boosting new physics searches with the help of an unsupervised anomaly detection algorithm. Prior to a search, each input event is preprocessed by the algorithm - a variational autoencoder (VAE). Based on the loss assigned to each event, input data can be split into a background control sample and a signal enriched sample. Following this strategy, one can enhance the sensitivity to new physics with no assumption on the underlying new physics signature. Our results show that a typical BSM search on the signal enriched group is more sensitive than an equivalent search on the original dataset
FPGA-accelerated machine learning inference as a service for particle physics computing
New heterogeneous computing paradigms on dedicated hardware with increased
parallelization, such as Field Programmable Gate Arrays (FPGAs), offer exciting
solutions with large potential gains. The growing applications of machine
learning algorithms in particle physics for simulation, reconstruction, and
analysis are naturally deployed on such platforms. We demonstrate that the
acceleration of machine learning inference as a web service represents a
heterogeneous computing solution for particle physics experiments that
potentially requires minimal modification to the current computing model. As
examples, we retrain the ResNet-50 convolutional neural network to demonstrate
state-of-the-art performance for top quark jet tagging at the LHC and apply a
ResNet-50 model with transfer learning for neutrino event classification. Using
Project Brainwave by Microsoft to accelerate the ResNet-50 image classification
model, we achieve average inference times of 60 (10) milliseconds with our
experimental physics software framework using Brainwave as a cloud (edge or
on-premises) service, representing an improvement by a factor of approximately
30 (175) in model inference latency over traditional CPU inference in current
experimental hardware. A single FPGA service accessed by many CPUs achieves a
throughput of 600--700 inferences per second using an image batch of one,
comparable to large batch-size GPU throughput and significantly better than
small batch-size GPU throughput. Deployed as an edge or cloud service for the
particle physics computing model, coprocessor accelerators can have a higher
duty cycle and are potentially much more cost-effective.Comment: 16 pages, 14 figures, 2 table
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors
Although the quest for more accurate solutions is pushing deep learning
research towards larger and more complex algorithms, edge devices demand
efficient inference and therefore reduction in model size, latency and energy
consumption. One technique to limit model size is quantization, which implies
using fewer bits to represent weights and biases. Such an approach usually
results in a decline in performance. Here, we introduce a method for designing
optimally heterogeneously quantized versions of deep neural network models for
minimum-energy, high-accuracy, nanosecond inference and fully automated
deployment on chip. With a per-layer, per-parameter type automatic quantization
procedure, sampling from a wide range of quantizers, model energy consumption
and size are minimized while high accuracy is maintained. This is crucial for
the event selection procedure in proton-proton collisions at the CERN Large
Hadron Collider, where resources are strictly limited and a latency of
s is required. Nanosecond inference and a resource
consumption reduced by a factor of 50 when implemented on field-programmable
gate array hardware are achieved
Fast convolutional neural networks on FPGAs with hls4ml
We introduce an automated tool for deploying ultra low-latency, low-power
deep neural networks with convolutional layers on FPGAs. By extending the
hls4ml library, we demonstrate an inference latency of s using
convolutional architectures, targeting microsecond latency applications like
those at the CERN Large Hadron Collider. Considering benchmark models trained
on the Street View House Numbers Dataset, we demonstrate various methods for
model compression in order to fit the computational constraints of a typical
FPGA device used in trigger and data acquisition systems of particle detectors.
In particular, we discuss pruning and quantization-aware training, and
demonstrate how resource utilization can be significantly reduced with little
to no loss in model accuracy. We show that the FPGA critical resource
consumption can be reduced by 97% with zero loss in model accuracy, and by 99%
when tolerating a 6% accuracy degradation.Comment: 18 pages, 18 figures, 4 table
Search for supersymmetry in pp collisions at = 13 tev in the single-lepton final state using the sum of masses of large-radius jets
Results are reported from a search for supersymmetric particles in proton-proton collisions in the final state with a single lepton, multiple jets, including at least one b-tagged jet, and large missing transverse momentum. The search uses a sample of proton-proton collision data at = 13 TeV recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 35.9 . The observed event yields in the signal regions are consistent with those expected from standard model backgrounds. The results are interpreted in the context of simplified models of supersymmetry involving gluino pair production, with gluino decay into either on- or off-mass-shell top squarks. Assuming that the top squarks decay into a top quark plus a stable, weakly interacting neutralino, scenarios with gluino masses up to about 1.9 TeV are excluded at 95% confidence level for neutralino masses up to about 1 TeV
Search for dark matter produced in association with heavy-flavor quark pairs in proton-proton collisions at = 13 TeV
A search is presented for an excess of events with heavy-flavor quark pairs ( and ) and a large imbalance in transverse momentum in data from proton–proton collisions at a center-of-mass energy of 13TeV. The data correspond to an integrated luminosity of 2.2 collected with the CMS detector at the CERN LHC. No deviations are observed with respect to standard model predictions. The results are used in the first interpretation of dark matter production in and final states in a simplified model. This analysis is also the first to perform a statistical combination of searches for dark matter produced with different heavy-flavor final states. The combination provides exclusions that are stronger than those achieved with individual heavy-flavor final states
Combination of searches for heavy resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collisions at = 8 and 13 TeV
A statistical combination of searches is presented for massive resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collision data collected by the CMS experiment at the LHC. The data were taken at centre-of-mass energies of 8 and 13 TeV, corresponding to respective integrated luminosities of 19.7 and up to 2.7 . The results are interpreted in the context of heavy vector triplet and singlet models that mimic properties of composite-Higgs models predicting W′ and Z′ bosons decaying to WZ, WW, WH, and ZH bosons. A model with a bulk graviton that decays into WW and ZZ is also considered. This is the first combined search for WW, WZ, WH, and ZH resonances and yields lower limits on masses at 95% confidence level for W′ and Z′ singlets at 2.3 TeV, and for a triplet at 2.4 TeV. The limits on the production cross section of a narrow bulk graviton resonance with the curvature scale of the warped extra dimension , in the mass range of 0.6 to 4.0 TeV, are the most stringent published to date
The CMS trigger system
This paper describes the CMS trigger system and its performance during Run 1 of the LHC. The trigger system consists of two levels designed to select events of potential physics interest from a GHz (MHz) interaction rate of proton-proton (heavy ion) collisions. The first level of the trigger is implemented in hardware, and selects events containing detector signals consistent with an electron, photon, muon, τ lepton, jet, or missing transverse energy. A programmable menu of up to 128 object-based algorithms is used to select events for subsequent processing. The trigger thresholds are adjusted to the LHC instantaneous luminosity during data taking in order to restrict the output rate to 100 kHz, the upper limit imposed by the CMS readout electronics. The second level, implemented in software, further refines the purity of the output stream, selecting an average rate of 400 Hz for offline event storage. The objectives, strategy and performance of the trigger system during the LHC Run 1 are described
- …